Skip to main content

All Questions

Tagged with
0votes
1answer
288views

Can transformer models be used to convert code from one programming language to another?

There was a question like this in 2019. I hope things have changed since then. Concretely, I am looking for a way to train a transformer model to convert code from SAS to Python. I guess the method ...
Vladimir's user avatar
3votes
1answer
1kviews

Creating a support chat bot for my business

I am trying to create a kind of support bot to answer questions from my clients about specific technical details about WordPress plugins that I sell. The goal is that the ...
digitalzoomstudio's user avatar
1vote
1answer
443views

OpeanAI Gym. Train problem: invalid values [closed]

I have a problem with my reinforcement learning model. I am trying to simulate an electric battery storage. To keep it as simple as possible, the efficiency of charge, storage and discharge are 100%. ...
MiPre's user avatar
1vote
0answers
130views

How should I simulate this Markov Decision Process?

I am working on solving a problem on nodes in a graph communicating with each other. They try to estimate a central state using Kalman consensus filter, with the connections described by the graph's ...
Suraj Joshi's user avatar
1vote
0answers
289views

How do i start building an autoclick bot for pubg mobile?

I want to make a bot which clicks the fire button on the mobile screen upon seeing an enemies head. In pubg mobile which is an android game you have to control the fire button and the aim along with ...
Fardeen Khan's user avatar
1vote
0answers
124views

Same implementation, but agent is not learning in Retro Pong Environment

I tried to implement the exact same python coding by Andrej Karpathy to train RL agent to play Pong, except that I migrated the environment from Gym to Retro. Everything is the same except the action ...
 Hoo's user avatar
1vote
0answers
201views

Understanding policy update in PPO2

I have a question regarding the functionality of the PPO2 algorithm together with the Stable Baselines implementation: From the original paper I know that the policy parameters $\theta$ are updated K-...
mglss's user avatar
1vote
1answer
767views

Advice on creating a new environment using OpenAI Gym [closed]

I'm looking for some general advice here before I dive in. I'm interested in creating a new environment for OpenAI gym to provide some slightly more challenging continuous control problems than the ...
Bill's user avatar
  • 141
1vote
2answers
323views

Running 2 NEAT nets on the same observations

So i have been playing around with neat-python. I made a program, applying neat, to play pinball on the Atari 2600. The code for that can be found in the file ...
niallmandal's user avatar
7votes
1answer
3kviews

2 Player Games in OpenAI Retro

I have been using OpenAI Retro for awhile, and I wanted to experiment with two player games. By two player games, I mean co-op games like "Tennis-Atari2600" or even Pong, where 2 agents are present in ...
niallmandal's user avatar
2votes
0answers
638views

How many episodes does it take for a vanilla one-step actor-critic agent to master the OpenAI BipedalWalker-v2 problem?

I'm trying to solve the OpenAI BipedalWalker-v2 by using a one-step actor-critic agent. I'm implementing the solution using python and tensorflow. I'm following this pseudo-code taken from the book ...
Diego Orellana's user avatar

close